A Generic Approach for Escaping Saddle points

نویسندگان

  • Sashank J. Reddi
  • Manzil Zaheer
  • Suvrit Sra
  • Barnabás Póczos
  • Francis R. Bach
  • Ruslan Salakhutdinov
  • Alexander J. Smola
چکیده

A central challenge to using first-order methods for optimizing nonconvex problems is the presence of saddle points. First-order methods often get stuck at saddle points, greatly deteriorating their performance. Typically, to escape from saddles one has to use second-order methods. However, most works on second-order methods rely extensively on expensive Hessian-based computations, making them impractical in large-scale settings. To tackle this challenge, we introduce a generic framework that minimizes Hessian based computations while at the same time provably converging to second-order critical points. Our framework carefully alternates between a firstorder and a second-order subroutine, using the latter only close to saddle points, and yields convergence results competitive to the state-of-the-art. Empirical results suggest that our strategy also enjoys good practical performance.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficient approaches for escaping higher order saddle points in non-convex optimization

Local search heuristics for non-convex optimizations are popular in applied machine learning. However, in general it is hard to guarantee that such algorithms even converge to a local minimum, due to the existence of complicated saddle point structures in high dimensions. Many functions have degenerate saddle points such that the first and second order derivatives cannot distinguish them with l...

متن کامل

Charged Point Normalization: An Efficient Solution to the Saddle Point Problem

Recently, the problem of local minima in very high dimensional non-convex optimization has been challenged and the problem of saddle points has been introduced. This paper introduces a dynamic type of normalization that forces the system to escape saddle points. Unlike other saddle point escaping algorithms, second order information is not utilized, and the system can be trained with an arbitra...

متن کامل

Escaping From Saddle Points - Online Stochastic Gradient for Tensor Decomposition

We analyze stochastic gradient descent for optimizing non-convex functions. In many cases for non-convex functions the goal is to find a reasonable local minimum, and the main concern is that gradient updates are trapped in saddle points. In this paper we identify strict saddle property for non-convex problem that allows for efficient optimization. Using this property we show that from an arbit...

متن کامل

Bowl breakout: escaping the positive region when searching for saddle points.

We present a scheme improving the minimum-mode following method for finding first order saddle points by confining the displacements of atoms to the subset of those subject to the largest force. By doing so it is ensured that the displacement remains of a local character within regions where all eigenvalues of the Hessian matrix are positive. However, as soon as a region is entered where an eig...

متن کامل

Analytical And Numerical Study Of Kramers ’ Exit Problem II ∗

Kramers’ exit problem is concerned with noise activated escape from a potential well. In case the noise strength, , (temperature measured in units of potential barrier height) is small this becomes a singular perturbation problem. It is shown that actually, most of the probability of the exit points on the separatrix is located at a distance O( √ ) from the saddle point and the probability vani...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1709.01434  شماره 

صفحات  -

تاریخ انتشار 2017